Optimal causal inference: estimating stored information and approximating causal architecture.
نویسندگان
چکیده
We introduce an approach to inferring the causal architecture of stochastic dynamical systems that extends rate-distortion theory to use causal shielding--a natural principle of learning. We study two distinct cases of causal inference: optimal causal filtering and optimal causal estimation. Filtering corresponds to the ideal case in which the probability distribution of measurement sequences is known, giving a principled method to approximate a system's causal structure at a desired level of representation. We show that in the limit in which a model-complexity constraint is relaxed, filtering finds the exact causal architecture of a stochastic dynamical system, known as the causal-state partition. From this, one can estimate the amount of historical information the process stores. More generally, causal filtering finds a graded model-complexity hierarchy of approximations to the causal architecture. Abrupt changes in the hierarchy, as a function of approximation, capture distinct scales of structural organization. For nonideal cases with finite data, we show how the correct number of the underlying causal states can be found by optimal causal estimation. A previously derived model-complexity control term allows us to correct for the effect of statistical fluctuations in probability estimates and thereby avoid overfitting.
منابع مشابه
Optimally Predictive Causal Inference
Natural systems compute intrinsically and produce information. The organization of a stochastic dynamical system is reflected in the time series of observations made of the system and can be quantified by the excess entropy or predictive information—the mutual information between past and future. This information can be used to build models of varying complexity that capture the causal structur...
متن کاملOptimal Causal Inference
We consider an information-theoretic objective function for statistical modeling of time series that embodies a parametrized trade-off between the predictive power of a model and the model’s complexity. We study two distinct cases of optimal causal inference, which we call optimal causal filtering (OCF) and optimal causal estimation (OCE). OCF corresponds to the ideal case of having infinite da...
متن کاملBayesian Nonparametric Causal Inference: Information Rates and Learning Algorithms
We investigate the problem of estimating the causal effect of a treatment on individual subjects from observational data; this is a central problem in various application domains, including healthcare, social sciences, and online advertising. Within the Neyman-Rubin potential outcomes model, we use the Kullback-Leibler (KL) divergence between the estimated and true distributions as a measure of...
متن کاملEstimating Directed Information to Infer Causal Relationships between Neural Spike Trains and Approximating Discrete Probability Distributions with Causal Dependence Trees by Christopher
This work examines an information theoretic quantity known as directed information, which measures statistically causal influences between processes. It is shown to be a general quantity, applicable to arbitrary probability distributions. It is interpreted in terms of prediction, communication with feedback, source coding with feed forward, control over noisy channels, and other settings. It is...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Chaos
دوره 20 3 شماره
صفحات -
تاریخ انتشار 2010